192 research outputs found

    A biologically inspired computational vision front-end based on a self-organised pseudo-randomly tessellated artificial retina

    Get PDF
    This paper considers the construction of a biologically inspired front-end for computer vision based on an artificial retina pyramid with a self-organised pseudo-randomly tessellated receptive field tessellation. The organisation of photoreceptors and receptive fields in biological retinae locally resembles a hexagonal mosaic, whereas globally these are organised with a very densely tessellated central foveal region which seamlessly merges into an increasingly sparsely tessellated periphery. In contrast, conventional computer vision approaches use a rectilinear sampling tessellation which samples the whole field of view with uniform density. Scale-space interest points which are suitable for higher level attention and reasoning tasks are efficiently extracted by our vision front-end by performing hierarchical feature extraction on the pseudo-randomly spaced visual information. All operations were conducted on a geometrically irregular foveated representation (data structure for visual information) which is radically different to the uniform rectilinear arrays used in conventional computer vision

    A topological approach for segmenting human body shape

    Get PDF
    Segmentation of a 3D human body, is a very challenging problem in applications exploiting human scan data. To tackle this problem, the paper proposes a topological approach based on the discrete Reeb graph (DRG) which is an extension of the classical Reeb graph to handle unorganized clouds of 3D points. The essence of the approach concerns detecting critical nodes in the DRG, thereby permitting the extraction of branches that represent parts of the body. Because the human body shape representation is built upon global topological features that are preserved so long as the whole structure of the human body does not change, our approach is quite robust against noise, holes, irregular sampling, frame change and posture variation. Experimental results performed on real scan data demonstrate the validity of our method

    Learning Rigid Image Registration - Utilizing Convolutional Neural Networks for Medical Image Registration

    Get PDF
    Many traditional computer vision tasks, such as segmentation, have seen large step-changes in accuracy and/or speed with the application of Convolutional Neural Networks (CNNs). Image registration, the alignment of two or more images to a common space, is a fundamental step in many medical imaging workflows. In this paper we investigate whether these techniques can also bring tangible benefits to the registration task. We describe and evaluate the use of convolutional neural networks (CNNs) for both mono- and multi- modality registration and compare their performance to more traditional schemes, namely multi-scale, iterative registration. This paper also investigates incorporating inverse consistency of the learned spatial transformations to impose additional constraints on the network during training and investigate any benefit in accuracy during detection. The approaches are validated with a series of artificial mono-modal registration tasks utilizing T1-weighted MR brain i mages from the Open Access Series of Imaging Studies (OASIS) study and IXI brain development dataset and a series of real multi-modality registration tasks using T1-weighted and T2-weighted MR brain images from the 2015 Ischemia Stroke Lesion segmentation (ISLES) challenge. The results demonstrate that CNNs give excellent performance for both mono- and multi- modality head and neck registration compared to the baseline method with significantly fewer outliers and lower mean errors

    A discrete Reeb graph approach for the segmentation of human body scans

    Get PDF
    Segmentation of 3D human body (HB) scan is a very challenging problem in applications exploiting human scan data. To tackle this problem, we propose a topological approach based on discrete Reeb graph (DRG) which is an extension of the classical Reeb graph to unorganized cloud of 3D points. The essence of the approach is detecting critical nodes in the DRG thus permitting the extraction of branches that represent the body parts. Because the human body shape representation is built upon global topological features that are preserved so long as the whole structure of the human body does not change, our approach is quite robust against noise, holes, irregular sampling, moderate reference change and posture variation. Experimental results performed on real scan data demonstrate the validity of our method

    Towards binocular active vision in a robot head system

    Get PDF
    This paper presents the first results of an investigation and pilot study into an active, binocular vision system that combines binocular vergence, object recognition and attention control in a unified framework. The prototype developed is capable of identifying, targeting, verging on and recognizing objects in a highly-cluttered scene without the need for calibration or other knowledge of the camera geometry. This is achieved by implementing all image analysis in a symbolic space without creating explicit pixel-space maps. The system structure is based on the ‘searchlight metaphor’ of biological systems. We present results of a first pilot investigation that yield a maximum vergence error of 6.4 pixels, while seven of nine known objects were recognized in a high-cluttered environment. Finally a “stepping stone” visual search strategy was demonstrated, taking a total of 40 saccades to find two known objects in the workspace, neither of which appeared simultaneously within the Field of View resulting from any individual saccade

    Self-correction of 3D reconstruction from multi-view stereo images

    Get PDF
    We present a self-correction approach to improving the 3D reconstruction of a multi-view 3D photogrammetry system. The self-correction approach has been able to repair the reconstructed 3D surface damaged by depth discontinuities. Due to self-occlusion, multi-view range images have to be acquired and integrated into a watertight nonredundant mesh model in order to cover the extended surface of an imaged object. The integrated surface often suffers from “dent” artifacts produced by depth discontinuities in the multi-view range images. In this paper we propose a novel approach to correcting the 3D integrated surface such that the dent artifacts can be repaired automatically. We show examples of 3D reconstruction to demonstrate the improvement that can be achieved by the self-correction approach. This self-correction approach can be extended to integrate range images obtained from alternative range capture devices

    Applying mesh conformation on shape analysis with missing data

    Get PDF
    A mesh conformation approach that makes use of deformable generic meshes has been applied to establishing correspondences between 3D shapes with missing data. Given a group of shapes with correspondences, we can build up a statistical shape model by applying principal component analysis (PCA). The conformation at first globally maps the generic mesh to the 3D shape based on manually located corresponding landmarks, and then locally deforms the generic mesh to clone the 3D shape. The local deformation is constrained by minimizing the energy of an elastic model. An algorithm was also embedded in the conformation process to fill missing surface data of the shapes. Using synthetic data, we demonstrate that the conformation preserves the configuration of the generic mesh and hence it helps to establish good correspondences for shape analysis. Case studies of the principal component analysis of shapes were presented to illustrate the successes and advantages of our approach

    Autonomous clothes manipulation using a hierarchical vision architecture

    Get PDF
    This paper presents a novel robot vision architecture for perceiving generic 3-D clothes configurations. Our architecture is hierarchically structured, starting from low-level curvature features to mid-level geometric shapes and topology descriptions, and finally, high-level semantic surface descriptions. We demonstrate our robot vision architecture in a customized dual-arm industrial robot with our inhouse developed stereo vision system, carrying out autonomous grasping and dual-arm flattening. The experimental results show the effectiveness of the proposed dual-arm flattening using the stereo vision system compared with the single-arm flattening using the widely cited Kinect-like sensor as the baseline. In addition, the proposed grasping approach achieves satisfactory performance when grasping various kind of garments, verifying the capability of the proposed visual perception architecture to be adapted to more than one clothing manipulation tasks

    Continuous perception for deformable objects understanding

    Get PDF
    We present a robot vision approach to deformable object classification, with direct application to autonomous service robots. Our approach is based on the assumption that continuous perception provides robots with greater visual competence for deformable objects interpretation and classification. Our approach thus classifies the category of clothing items by continuously perceiving the dynamic interactions of the garment’s material and shape as it is being picked up. Our proposed solution consists of extracting continuously visual features of a RGB-D video sequence and fusing features by means of the Locality Constrained Group Sparse Representation (LGSR) algorithm. To evaluate the performance of our approach, we created a fully annotated database featuring 150 garment videos in random configurations. Experiments demonstrate that by continuously observing an object deform, our approach achieves a classification score of 66.7%, outperforming state-of-the-art approaches by a ∼ 27.3% increase

    Integration of range images in a multi-view stereo system

    Full text link
    A novel method for integrating multiple range images in a multi-view stereo imaging system is presented here. Due to self-occlusion an individual range image provides only a partial model of an object surface. Therefore multiple range images from differing viewpoints must be captured and merged to extend the surface area that can be captured. In our approach range images are decomposed into subset patches and then evaluated in a "confidence competition". Redundant patches are removed whilst winning patches are merged to complete a single plausible mesh that represents the acquired object surface
    corecore